8 research outputs found
Commissioning of the CMS High Level Trigger
The CMS experiment will collect data from the proton-proton collisions
delivered by the Large Hadron Collider (LHC) at a centre-of-mass energy up to
14 TeV. The CMS trigger system is designed to cope with unprecedented
luminosities and LHC bunch-crossing rates up to 40 MHz. The unique CMS trigger
architecture only employs two trigger levels. The Level-1 trigger is
implemented using custom electronics, while the High Level Trigger (HLT) is
based on software algorithms running on a large cluster of commercial
processors, the Event Filter Farm. We present the major functionalities of the
CMS High Level Trigger system as of the starting of LHC beams operations in
September 2008. The validation of the HLT system in the online environment with
Monte Carlo simulated data and its commissioning during cosmic rays data taking
campaigns are discussed in detail. We conclude with the description of the HLT
operations with the first circulating LHC beams before the incident occurred
the 19th September 2008
The CMS event builder and storage system.
The CMS event builder assembles events accepted by the first level trigger and makes them available to the high-level trigger. The event builder needs to handle a maximum input rate of 100\,kHz and an aggregated throughput of 100\,GB/s originating from approximately 500 sources. This paper presents the chosen hardware and software architecture. The system consists of 2 stages: an initial pre-assembly reducing the number of fragments by one order of magnitude and a final assembly by several independent readout builder (RU-builder) slices. The RU-builder is based on 3 separate services: the buffering of event fragments during the assembly, the event assembly, and the data flow manager. A further component is responsible for handling events accepted by the high-level trigger: the storage manager (SM) temporarily stores the events on disk at a peak rate of 2\,GB/s until they are permanently archived offline. In addition, events and data-quality histograms are served by the SM to online monitoring clients. We discuss the operational experience from the first months of reading out cosmic ray data with the complete CMS detector
Dynamic configuration of the CMS data Acquisition cluster
The CMS online cluster consists of more than 2000 computers running about 10000 application instances. These applications implement the control of the experiment, the event building, the high level trigger, the online database and the control of the buffering and transferring of data to the Central Data Recording at CERN. In this paper the IT solutions employed to fulfil the requirements of such a large cluster are revised. Details are given on the chosen network structure, configuration management system, monitoring infrastructure and on the implementation of the high availability for the services and infrastructure